Asymptotic Properties of Nearest Neighbor Rules Using Edited Data
نویسنده
چکیده
The convergence properties of a nearest neighbor rule that uses an editing procedure to reduce the number of preclassified samples and to improve the performance of the rule are developed. Editing of the preclassified samples using the three-nearest neighbor rule followed by classification using the single-nearest neighbor rule with the remaining preclassified samples appears to produce a decision procedure whose risk approaches the Bayes' risk quite closely in many problems with only a few preclassified samples. The asymptotic risk of the nearest neighbor rules and the nearest neighbor rules using edited preclassified samples is calculated for several problems.
منابع مشابه
Asymptotic Behaviors of Nearest Neighbor Kernel Density Estimator in Left-truncated Data
Kernel density estimators are the basic tools for density estimation in non-parametric statistics. The k-nearest neighbor kernel estimators represent a special form of kernel density estimators, in which the bandwidth is varied depending on the location of the sample points. In this paper, we initially introduce the k-nearest neighbor kernel density estimator in the random left-truncatio...
متن کاملConsiderations about sample-size sensitivity of a family of edited nearest-neighbor rules
The edited nearest neighbor classification rules constitute a valid alternative to k-NN rules and other nonparametric classifiers. Experimental results with synthetic and real data from various domains and from different researchers and practitioners suggest that some editing algorithms (especially, the optimal ones) are very sensitive to the total number of prototypes considered. This paper in...
متن کاملSubmitted to CVPR ' 99 Bayesian based Optimal Nearest Neighbor
Nearest neighbor rules are popular classiiers, partly due to their good asymptotic properties. Improving the performance of NN rules when only nite samples are available has been studied over the last two decades. For example, smart distance measures which are data dependent have been proposed. From a theoretical point of view, one is interested in deriving the optimal NN distance measure. Exis...
متن کاملOn the edited fuzzy K-nearest neighbor rule
Classification of objects is an important area in a variety of fields and applications. In the presence of full knowledge of the underlying joint distributions, Bayes analysis yields an optimal decision procedure and produces optimal error rates. Many different methods are available to make a decision in those cases where information of the underlying joint distributions is not presented. The k...
متن کاملApplication of Proximity Graphs to Editing Nearest Neighbor Decision Rules
Non-parametric decision rules, such as the nearest neighbor (NN) rule, are attractive because no a priori knowledge is required concerning the underlying distributions of the data. Two traditional criticisms directed at the NN-rule concern the large amounts of storage and computation involved due to the apparent necessity to store all the sample (training) data. Thus there has been considerable...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- IEEE Trans. Systems, Man, and Cybernetics
دوره 2 شماره
صفحات -
تاریخ انتشار 1972